Model Generator and Deep Learning Architectures

The Model Generator dialog for the Deep Learning Tool and the Segmentation Wizard provides settings for generating new deep models with a number of different architectures, as well as for downloading pre-trained models (see Pre-Trained Models).

Click the New button on the Model Overview panel to open the Model Generator dialog, shown below.

Model Generator dialog

The following table lists the settings that are available for generating semantic segmentation, super-resolution, and denoising deep models.

Model Generator settings
  Description
Show architectures for Lets you filter the available architectures to those recommended for segmentation, super-resolution, and denoising.

Semantic Segmentation… Filters the Architecture list to models best suited for semantic segmentation, which is the process of associating each pixel of an image with a class label, such as a material phase or anatomical feature. Semantic segmentation models are suitable for binary and multi-class semantic segmentation tasks.

Super-resolution… Filters the Architecture list to models best suited for super-resolution.

Denoising… Filters the Architecture list to models best suited for denoising.

Architecture Lists the default models supplied with the Deep Learning Tool and the Segmentation Wizard. Architectures can be filtered by type (see Architectures and Editable Parameters for Semantic Segmentation Models and Architectures and Editable Parameters for Regression Models).

Note You can also download a selection of pre-trained models (see Pre-Trained Models).

Architecture description Provides a short description of the selected architecture and a link for further information.
Model type Lets you choose the type of model — Regression or Semantic segmentation — that you need to generate.
Class count Available only for semantic segmentation model types, this parameter lets you enter the number of classes required. The minimum number of classes is '2', which would be for a binary segmentation task, while the maximum is '20' for multi-class segmentations.
Input count Lets you include multiple inputs for training. For example, when you are working with data from simultaneous image acquisition systems you might want to select each modality as an input.

If additional inputs are added, then you can name them in the 'Input names' edit boxes.

Input dimension Lets you select an input dimension, as follows.

2D… The 2D approach analyzes and infers one slice of the image at a time. All feature maps and parameter tensors in all layers are 2D.

2.5D… The 2.5D approach analyzes a number of consecutive slices of the image as input channels. The remaining parts of the 2.5D model, including the feature maps and parameter tensors, are 2D, and the model output is the inferred 2D middle slice within the selected number of slices. You can choose the number of slices, as shown below.

3D… The 3D approach analyzes and infers the image volume in three-dimensional space. This approach is often more reliable for inference when compared to 2D and 2.5D approaches. but may need more computational memory to train and apply.

Note Only U-Net 3D is a true 3D model that uses 3D convolutions. The number of input slices for this model is determined by the input size, which must be cubic. For example, 32x32x32. U-Net uses 2D convolutions, but can take 2.5D input patches for which you can choose the number of slices. You should also note that in some cases, 3D models can be more reliable for segmentation tasks.

Name Lets you enter a name for the generated model.
Description Lets you enter a description of your model.
Parameters Lists the hyperparameters associated with the selected architecture and the default values for each (see Architectures and Editable Parameters for Semantic Segmentation Models and Architectures and Editable Parameters for Regression Models.